dictionary approach
A Dictionary Approach to Domain-Invariant Learning in Deep Networks
In this paper, we consider domain-invariant deep learning by explicitly modeling domain shifts with only a small amount of domain-specific parameters in a Convolutional Neural Network (CNN). By exploiting the observation that a convolutional filter can be well approximated as a linear combination of a small set of dictionary atoms, we show for the first time, both empirically and theoretically, that domain shifts can be effectively handled by decomposing a convolutional layer into a domain-specific atom layer and a domain-shared coefficient layer, while both remain convolutional. An input channel will now first convolve spatially only with each respective domain-specific dictionary atom to ``absorb domain variations, and then output channels are linearly combined using common decomposition coefficients trained to promote shared semantics across domains. We use toy examples, rigorous analysis, and real-world examples with diverse datasets and architectures, to show the proposed plug-in framework's effectiveness in cross and joint domain performance and domain adaptation. With the proposed architecture, we need only a small set of dictionary atoms to model each additional domain, which brings a negligible amount of additional parameters, typically a few hundred.
Review for NeurIPS paper: A Dictionary Approach to Domain-Invariant Learning in Deep Networks
Additional Feedback: Some issues need to be addressed: 1) Missing baselines. Recent studies [1-2] have shown that domain-specific BN may improve the adaptation performance. So why in Basic Branching and DAFD, all BNs are shared across domains? Maybe for DAFD, it is reasonable because it assumes filter decomposition could'correct' the shifts. But for the Basic Branching, not sharing BN across domains may be helpful.
A Dictionary Approach to Domain-Invariant Learning in Deep Networks
In this paper, we consider domain-invariant deep learning by explicitly modeling domain shifts with only a small amount of domain-specific parameters in a Convolutional Neural Network (CNN). By exploiting the observation that a convolutional filter can be well approximated as a linear combination of a small set of dictionary atoms, we show for the first time, both empirically and theoretically, that domain shifts can be effectively handled by decomposing a convolutional layer into a domain-specific atom layer and a domain-shared coefficient layer, while both remain convolutional. An input channel will now first convolve spatially only with each respective domain-specific dictionary atom to absorb" domain variations, and then output channels are linearly combined using common decomposition coefficients trained to promote shared semantics across domains. We use toy examples, rigorous analysis, and real-world examples with diverse datasets and architectures, to show the proposed plug-in framework's effectiveness in cross and joint domain performance and domain adaptation. With the proposed architecture, we need only a small set of dictionary atoms to model each additional domain, which brings a negligible amount of additional parameters, typically a few hundred.
Modeling and Forecasting COVID-19 Cases using Latent Subpopulations
Vega, Roberto, Shah, Zehra, Ramazi, Pouria, Greiner, Russell
Classical epidemiological models assume homogeneous populations. There have been important extensions to model heterogeneous populations, when the identity of the sub-populations is known, such as age group or geographical location. Here, we propose two new methods to model the number of people infected with COVID-19 over time, each as a linear combination of latent sub-populations -- i.e., when we do not know which person is in which sub-population, and the only available observations are the aggregates across all sub-populations. Method #1 is a dictionary-based approach, which begins with a large number of pre-defined sub-population models (each with its own starting time, shape, etc), then determines the (positive) weight of small (learned) number of sub-populations. Method #2 is a mixture-of-$M$ fittable curves, where $M$, the number of sub-populations to use, is given by the user. Both methods are compatible with any parametric model; here we demonstrate their use with first (a)~Gaussian curves and then (b)~SIR trajectories. We empirically show the performance of the proposed methods, first in (i) modeling the observed data and then in (ii) forecasting the number of infected people 1 to 4 weeks in advance. Across 187 countries, we show that the dictionary approach had the lowest mean absolute percentage error and also the lowest variance when compared with classical SIR models and moreover, it was a strong baseline that outperforms many of the models developed for COVID-19 forecasting.
- North America > Canada > Alberta (0.15)
- North America > United States (0.14)
- Africa > Middle East > Egypt (0.04)
- (5 more...)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (1.00)
- Health & Medicine > Therapeutic Area > Immunology (1.00)
- Health & Medicine > Epidemiology (1.00)
Nonlinear Functional Output Regression: a Dictionary Approach
Bouche, Dimitri, Clausel, Marianne, Roueff, François, d'Alché-Buc, Florence
In a large number of fields such as biomedical signal processing, speech and acoustics and climate science, data consists of a high number of simultaneous or sequential measurements of different aspects of the same phenomenon. Such data is inherently high dimensional, however it contains strong within-observation correlations and smoothness patterns which can be utilized in the learning process. A possible way to do so is to represent those observations as functions rather than vectors, opening the door to Functional Data Analysis (FDA) Ramsay & Silverman (2005), a research area that has recently attracted a growing interest due to the ubiquity of embedded devices and sensor data. In practice, FDA relies on the assumption that the sampling rate at which data are collected is high enough to get functional observations. Of special interest is the general problem of functional-output regression in which the output variable to regress is a function and no specific assumption is made on the input variable that can can be of any type, including functions. While functional linear model have received a great deal of attention--see Ramsay & Silverman (2005), Morris (2015) and references therein--nonlinear ones have been less studied.
- North America > United States (0.54)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)